📚 node [[tpu_master|tpu master]]
Welcome! Nobody has contributed anything to 'tpu_master|tpu master' yet. You can:
  • Write something in the document below!
    • There is at least one public document in every node in the Agora. Whatever you write in it will be integrated and made available for the next visitor to read and edit.
  • Write to the Agora from social media.
    • If you follow Agora bot on a supported platform and include the wikilink [[tpu_master|tpu master]] in a post, the Agora will link it here and optionally integrate your writing.
  • Sign up as a full Agora user.
    • As a full user you will be able to contribute your personal notes and resources directly to this knowledge commons. Some setup required :)
⥅ related node [[tpu_master]]
⥅ node [[tpu_master]] pulled by Agora

TPU master

Go back to the [[AI Glossary]]

#TensorFlow #GoogleCloud

The central coordination process running on a host machine that sends and receives data, results, programs, performance, and system health information to the TPU workers. The TPU master also manages the setup and shutdown of TPU devices.

📖 stoas
⥱ context